2025-04-03 23:26:48,838 [ 233349 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-03 23:26:48,838 [ 233349 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-03 23:26:48,838 [ 233349 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-03 23:26:48,838 [ 233349 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_9hb1f2 --privileged --dns-search='.' --memory=30709030912 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=1 --color=no --durations=0 test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_backward_compatibility/test_functions.py::test_string_functions -vvv" altinityinfra/integration-tests-runner:cd6390247eca '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: random-0.2, timeout-2.2.0, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False collecting ... collected 6 items test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active SKIPPED [ 16%] test_attach_partition_using_copy/test.py::test_all_replicated FAILED [ 33%] test_attach_partition_using_copy/test.py::test_both_mergetree FAILED [ 50%] test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED [ 66%] test_attach_partition_using_copy/test.py::test_only_destination_replicated FAILED [ 83%] test_backward_compatibility/test_functions.py::test_string_functions SKIPPED [100%] =================================== FAILURES =================================== _____________________________ test_all_replicated ______________________________ start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:126: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2025-04-03 23:27:11 [ 625 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2793, start) 2025-04-03 23:27:11 [ 625 ] DEBUG : Cluster start called. is_up=False (cluster.py:2800, start) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-03 23:27:11 [ 625 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-03 23:27:11 [ 625 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy (cluster.py:922, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-03 23:27:11 [ 625 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-03 23:27:11 [ 625 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-03 23:27:11 [ 625 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-03 23:27:11 [ 625 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2813, start) 2025-04-03 23:27:11 [ 625 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/configs/config.d (cluster.py:4752, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/database (cluster.py:4769, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/logs (cluster.py:4780, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2813, start) 2025-04-03 23:27:11 [ 625 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/configs/config.d (cluster.py:4752, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/database (cluster.py:4769, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/logs (cluster.py:4780, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-03 23:27:11 [ 625 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env (cluster.py:97, _create_env_file) 2025-04-03 23:27:11 [ 625 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-03 23:27:11 [ 625 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-03 23:27:11 [ 625 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-03 23:27:11 [ 625 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-03 23:27:11 [ 625 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-03 23:27:11 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --project-name roottestattachpartitionusingcopy --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/docker-compose.yml pull] (cluster.py:122, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by replica1 (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: replica1 Pulling (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: replica1 Pulled (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Setup ZooKeeper (cluster.py:2854, start) 2025-04-03 23:27:22 [ 625 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/keeper3/coordination'] (cluster.py:2855, start) 2025-04-03 23:27:22 [ 625 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:122, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr:time="2025-04-03T23:27:22Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy_default Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy_default Created (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Created (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Created (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Created (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Starting (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Starting (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Starting (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Started (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Started (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Started (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr:time="2025-04-03T23:27:22Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Stderr:time="2025-04-03T23:27:22Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-03 23:27:22 [ 625 ] DEBUG : Wait ZooKeeper to start (cluster.py:2466, wait_zookeeper_to_start) 2025-04-03 23:27:22 [ 625 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2082, get_instance_ip) 2025-04-03 23:27:22 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:22 [ 625 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.1.3, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 23:27:22 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:22 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:22 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:22 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:22 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:22 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:23 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:23 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:23 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:23 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:25 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:25 [ 625 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] INFO : Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 23:27:28 [ 625 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2082, get_instance_ip) 2025-04-03 23:27:28 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:28 [ 625 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.1.4, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 23:27:28 [ 625 ] INFO : Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 23:27:28 [ 625 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2082, get_instance_ip) 2025-04-03 23:27:28 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:28 [ 625 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.1.2, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-03 23:27:28 [ 625 ] INFO : Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-03 23:27:28 [ 625 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-03 23:27:28 [ 625 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-03 23:27:28 [ 625 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-03 23:27:28 [ 625 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-03 23:27:28 [ 625 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2482, wait_zookeeper_nodes_to_start) 2025-04-03 23:27:28 [ 625 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --project-name roottestattachpartitionusingcopy --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3200, start) 2025-04-03 23:27:28 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --project-name roottestattachpartitionusingcopy --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:122, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Running (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Running (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Running (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Creating (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Created (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Created (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Starting (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Starting (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Started (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Started (cluster.py:148, run_and_check) 2025-04-03 23:27:29 [ 625 ] DEBUG : ClickHouse instance created (cluster.py:3208, start) 2025-04-03 23:27:29 [ 625 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2082, get_instance_ip) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2092, get_instance_global_ipv6) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.1.6... (cluster.py:3216, start) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:29 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/b508ad0d7182801153b6589ee694f6721aa17161e40f2fc12bba6ccd18912de7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : ClickHouse replica1 started (cluster.py:3220, start) 2025-04-03 23:27:30 [ 625 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2082, get_instance_ip) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2092, get_instance_global_ipv6) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.1.5... (cluster.py:3216, start) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : http://localhost:None "GET /v1.46/containers/9549ab9043435679486adb901fe61b1f87e4bfc5e7468806f102db2595b00683/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-03 23:27:30 [ 625 ] DEBUG : ClickHouse replica2 started (cluster.py:3220, start) ------------------------------ Captured log call ------------------------------- 2025-04-03 23:27:30 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:27:30 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:27:30 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:27:30 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:27:31 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:28:24 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:29:19 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:30:14 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:31:11 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:32:09 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:33:04 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:33:59 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:34:56 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:35:54 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:36:48 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:37:43 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:38:40 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:39:38 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:40:33 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:41:28 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _____________________________ test_both_mergetree ______________________________ start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:104: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-03 23:42:11 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:42:25 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:42:25 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:42:25 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:42:25 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:43:23 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:44:17 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:45:13 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:46:10 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:47:07 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:48:02 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:48:57 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:49:54 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:50:52 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:51:47 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:52:42 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:53:39 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:54:36 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:55:31 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:56:26 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_not_work_on_different_disk ________________________ start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:197: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-03 23:57:11 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:57:23 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-03 23:57:23 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:57:23 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-03 23:57:23 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:58:21 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-03 23:59:16 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:00:11 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:01:08 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:02:06 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:03:00 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:03:55 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:04:52 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:05:50 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:06:45 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:07:40 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:08:37 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:09:35 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:10:29 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:11:25 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_only_destination_replicated _______________________ start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-04 00:12:11 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-04 00:12:21 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-04 00:12:21 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-04 00:12:21 [ 625 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-04 00:12:22 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:13:19 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:14:14 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:15:09 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:16:06 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:17:02 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:17:58 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:18:52 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:19:47 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:20:44 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:21:42 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:22:37 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:23:32 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:24:29 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:25:27 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-04 00:26:21 [ 625 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) ---------------------------- Captured log teardown ----------------------------- 2025-04-04 00:27:12 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --project-name roottestattachpartitionusingcopy --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/docker-compose.yml stop --timeout 20] (cluster.py:122, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-04 00:27:24 [ 625 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/.env --project-name roottestattachpartitionusingcopy --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-1/replica2/docker-compose.yml down --volumes] (cluster.py:122, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica2-1 Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-replica1-1 Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo1-1 Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo3-1 Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-zoo2-1 Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy_default Removing (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy_default Removed (cluster.py:148, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-04 00:27:25 [ 625 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-04 00:27:25 [ 625 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-04 00:27:25 [ 625 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy (cluster.py:922, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-04 00:27:25 [ 625 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-04 00:27:25 [ 625 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) ============================== slowest durations =============================== 900.00s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 900.00s call test_attach_partition_using_copy/test.py::test_both_mergetree 900.00s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 881.07s call test_attach_partition_using_copy/test.py::test_all_replicated 18.93s setup test_attach_partition_using_copy/test.py::test_all_replicated 13.06s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 12.77s setup test_backward_compatibility/test_functions.py::test_string_functions 12.48s setup test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active 4.41s teardown test_backward_compatibility/test_functions.py::test_string_functions 4.10s teardown test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active 0.60s call test_backward_compatibility/test_functions.py::test_string_functions 0.17s call test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated =========================== short test summary info ============================ FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Failed... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Failed... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated SKIPPED [1] test_asynchronous_metric_jemalloc_profile_active/test.py:30: Disabled for sanitizers SKIPPED [1] test_backward_compatibility/test_functions.py:164: The test is slow in builds with sanitizer ================== 4 failed, 2 skipped in 3648.26s (1:00:48) =================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 528, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_9hb1f2 --privileged --dns-search='.' --memory=30709030912 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=1 --color=no --durations=0 test_asynchronous_metric_jemalloc_profile_active/test.py::test_asynchronous_metric_jemalloc_profile_active test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_backward_compatibility/test_functions.py::test_string_functions -vvv" altinityinfra/integration-tests-runner:cd6390247eca ' returned non-zero exit status 1.